An Approximate Fisher Scoring Algorithm for Finite Mixtures of Multinomials
نویسندگان
چکیده
Finite mixture distributions arise naturally in many applications including clustering and classification. Since they usually do not yield closed forms for maximum likelihood estimates (MLEs), numerical methods using the well known Fisher Scoring or Expectation-Maximization algorithms are considered. In this work, an approximation to the Fisher Information Matrix of an arbitrary mixture of multinomial distributions is introduced. This leads to an Approximate Fisher Scoring algorithm (AFSA), which turns out to be closely related to Expectation-Maximization, and is more robust to the choice of initial value than Fisher Scoring iterations. A combination of AFSA and the classical Fisher Scoring iterations provides the best of both computational efficiency and stable convergence properties.
منابع مشابه
Fisher Scoring and a Mixture of Modes Approach for Approximate Inference and Learning in Nonlinear State Space Models
We present Monte-Carlo generalized EM equations for learning in nonlinear state space models. The difficulties lie in the Monte-Carlo E-step which consists of sampling from the posterior distribution of the hidden variables given the observations. The new idea presented in this paper is to generate samples from a Gaussian approximation to the true posterior from which it is easy to obtain indep...
متن کاملmovMF: An R Package for Fitting Mixtures of von Mises-Fisher Distributions
Finite mixtures of von Mises-Fisher distributions allow to apply model-based clustering methods to data which is of standardized length, i.e., all data points lie on the unit sphere. The R package movMF contains functionality to draw samples from finite mixtures of von Mises-Fisher distributions and to fit these models using the expectation-maximization algorithm for maximum likelihood estimati...
متن کاملClustering in Hilbert simplex geometry
Clustering categorical distributions in the probability simplex is a fundamental primitive often met in applications dealing with histograms or mixtures of multinomials. Traditionally, the differentialgeometric structure of the probability simplex has been used either by (i) setting the Riemannian metric tensor to the Fisher information matrix of the categorical distributions, or (ii) defining ...
متن کاملCompare Adomian Decomposition Method and Laplace Decomposition Method for Burger's-Huxley and Burger's-Fisher equations
In this paper, Adomian decomposition method (ADM) and Laplace decomposition method (LDM) used to obtain series solutions of Burgers-Huxley and Burgers-Fisher Equations. In ADM the algorithm is illustrated by studying an initial value problem and LDM is based on the application of Laplace transform to nonlinear partial differential equations. In ADM only few terms of the expansion are required t...
متن کاملStochastic Gradient Descent as Approximate Bayesian Inference
Stochastic Gradient Descent with a constant learning rate (constant SGD) simulates a Markov chain with a stationary distribution. With this perspective, we derive several new results. (1) We show that constant SGD can be used as an approximate Bayesian posterior inference algorithm. Specifically, we show how to adjust the tuning parameters of constant SGD to best match the stationary distributi...
متن کامل